Goto

Collaborating Authors

 nonlinear ica







ICE-BeeM: IdentifiableConditionalEnergy-Based DeepModelsBasedonNonlinearICA

Neural Information Processing Systems

Our results extend recent developments innonlinear ICA, and in fact, they lead to an important generalization of ICA models. In particular, we show that our model can be used for the estimation of the components in theframeworkofIndependentlyModulatedComponentAnalysis(IMCA),anew generalization of nonlinear ICA that relaxes the independence assumption.


On the Identifiability of Nonlinear ICA: Sparsity and Beyond

Neural Information Processing Systems

Nonlinear independent component analysis (ICA) aims to recover the underlying independent latent sources from their observable nonlinear mixtures. How to make the nonlinear ICA model identifiable up to certain trivial indeterminacies is a long-standing problem in unsupervised learning. Recent breakthroughs reformulate the standard independence assumption of sources as conditional independence given some auxiliary variables (e.g., class labels and/or domain/time indexes) as weak supervision or inductive bias.


ICE-BeeM: Identifiable Conditional Energy-Based Deep Models Based on Nonlinear ICA

Neural Information Processing Systems

We consider the identifiability theory of probabilistic models and establish sufficient conditions under which the representations learnt by a very broad family of conditional energy-based models are unique in function space, up to a simple transformation. In our model family, the energy function is the dot-product between two feature extractors, one for the dependent variable, and one for the conditioning variable. We show that under mild conditions, the features are unique up to scaling and permutation. Our results extend recent developments in nonlinear ICA, and in fact, they lead to an important generalization of ICA models. In particular, we show that our model can be used for the estimation of the components in the framework of Independently Modulated Component Analysis (IMCA), a new generalization of nonlinear ICA that relaxes the independence assumption. A thorough empirical study show that representations learnt by our model from real-world image datasets are identifiable, and improve performance in transfer learning and semi-supervised learning tasks.